Goto

Collaborating Authors

 different objective


28e209b61a52482a0ae1cb9f5959c792-AuthorFeedback.pdf

Neural Information Processing Systems

We deeply appreciate the reviewers' careful comments. We hope all concerns can be resolved through our clarifications. Q: I'd recommend picking an OOD detection threshould at 95% TPR for a more even comparison. A: Thank you for your great suggestion. Previously we set the threshold at 0.5 as a default value for binary classification.


Uncertain Multi-Objective Recommendation via Orthogonal Meta-Learning Enhanced Bayesian Optimization

Wang, Hongxu, Sun, Zhu, Du, Yingpeng, Zhang, Lu, He, Tiantian, Ong, Yew-Soon

arXiv.org Artificial Intelligence

Recommender systems (RSs) play a crucial role in shaping our digital interactions, influencing how we access and engage with information across various domains. Traditional research has predominantly centered on maximizing recommendation accuracy, often leading to unintended side effects such as echo chambers and constrained user experiences. Drawing inspiration from autonomous driving, we introduce a novel framework that categorizes RS autonomy into five distinct levels, ranging from basic rule-based accuracy-driven systems to behavior-aware, uncertain multi-objective RSs - where users may have varying needs, such as accuracy, diversity, and fairness. In response, we propose an approach that dynamically identifies and optimizes multiple objectives based on individual user preferences, fostering more ethical and intelligent user-centric recommendations. To navigate the uncertainty inherent in multi-objective RSs, we develop a Bayesian optimization (BO) framework that captures personalized trade-offs between different objectives while accounting for their uncertain interdependencies. Furthermore, we introduce an orthogonal meta-learning paradigm to enhance BO efficiency and effectiveness by leveraging shared knowledge across similar tasks and mitigating conflicts among objectives through the discovery of orthogonal information. Finally, extensive empirical evaluations demonstrate the effectiveness of our method in optimizing uncertain multi-objectives for individual users, paving the way for more adaptive and user-focused RSs.


Multi-Objective Planning with Contextual Lexicographic Reward Preferences

Rustagi, Pulkit, Anand, Yashwanthi, Saisubramanian, Sandhya

arXiv.org Artificial Intelligence

Autonomous agents are often required to plan under multiple objectives whose preference ordering varies based on context. The agent may encounter multiple contexts during its course of operation, each imposing a distinct lexicographic ordering over the objectives, with potentially different reward functions associated with each context. Existing approaches to multi-objective planning typically consider a single preference ordering over the objectives, across the state space, and do not support planning under multiple objective orderings within an environment. We present Contextual Lexicographic Markov Decision Process (CLMDP), a framework that enables planning under varying lexicographic objective orderings, depending on the context. In a CLMDP, both the objective ordering at a state and the associated reward functions are determined by the context. We employ a Bayesian approach to infer a state-context mapping from expert trajectories. Our algorithm to solve a CLMDP first computes a policy for each objective ordering and then combines them into a single context-aware policy that is valid and cycle-free. The effectiveness of the proposed approach is evaluated in simulation and using a mobile robot.


MODRL-TA:A Multi-Objective Deep Reinforcement Learning Framework for Traffic Allocation in E-Commerce Search

Cheng, Peng, Wang, Huimu, Zhao, Jinyuan, Wang, Yihao, Xu, Enqiang, Zhao, Yu, Xiao, Zhuojian, Wang, Songlin, Tang, Guoyu, Liu, Lin, Xu, Sulong

arXiv.org Artificial Intelligence

Traffic allocation is a process of redistributing natural traffic to products by adjusting their positions in the post-search phase, aimed at effectively fostering merchant growth, precisely meeting customer demands, and ensuring the maximization of interests across various parties within e-commerce platforms. Existing methods based on learning to rank neglect the long-term value of traffic allocation, whereas approaches of reinforcement learning suffer from balancing multiple objectives and the difficulties of cold starts within realworld data environments. To address the aforementioned issues, this paper propose a multi-objective deep reinforcement learning framework consisting of multi-objective Q-learning (MOQ), a decision fusion algorithm (DFM) based on the cross-entropy method(CEM), and a progressive data augmentation system(PDA). Specifically. MOQ constructs ensemble RL models, each dedicated to an objective, such as click-through rate, conversion rate, etc. These models individually determine the position of items as actions, aiming to estimate the long-term value of multiple objectives from an individual perspective. Then we employ DFM to dynamically adjust weights among objectives to maximize long-term value, addressing temporal dynamics in objective preferences in e-commerce scenarios. Initially, PDA trained MOQ with simulated data from offline logs. As experiments progressed, it strategically integrated real user interaction data, ultimately replacing the simulated dataset to alleviate distributional shifts and the cold start problem. Experimental results on real-world online e-commerce systems demonstrate the significant improvements of MODRL-TA, and we have successfully deployed MODRL-TA on an e-commerce search platform.


VBMO: Voting-Based Multi-Objective Path Planning

Korpan, Raj

arXiv.org Artificial Intelligence

This paper presents VBMO, the Voting-Based Multi-Objective path planning algorithm, that generates optimal single-objective plans, evaluates each of them with respect to the other objectives, and selects one with a voting mechanism. VBMO does not use hand-tuned weights, consider the multiple objectives at every step of search, or use an evolutionary algorithm. Instead, it considers how a plan that is optimal in one objective may perform well with respect to others. VBMO incorporates three voting mechanisms: range, Borda, and combined approval. Extensive evaluation in diverse and complex environments demonstrates the algorithm's ability to efficiently produce plans that satisfy multiple objectives.


Multi-Objective Recommender System- A need of the hour

#artificialintelligence

My reading so far, is that designing and implementing a real time automated recommendation is a probem that requires multiple objectives to solve and this is the need of the hour. . Such multi-objective system needs competing objectives of consumers, possible tensions between goals of different stakeholders, conflicts when optimizing for different time horizons, competing design choices at the UI level, as well as system-level and engineering-related considerations. Solution I am using the @kaggle competation is kind of enesemble modle RNN, Transformers and Linear Programming.


End-to-end machine learning lifecycle

#artificialintelligence

A machine learning (ML) project requires collaboration across multiple roles in a business. We'll introduce the high level steps of what the end-to-end ML lifecycle looks like and how different roles can collaborate to complete the ML project. Machine learning is a powerful tool to help solve different problems in your business. The article "Building your first machine learning model" gives you basic ideas of what it takes to build a machine learning model. In this article, we'll talk about what the end-to-end machine learning project lifecycle looks like in a real business.


End-to-end machine learning lifecycle

#artificialintelligence

A machine learning (ML) project requires collaboration across multiple roles in a business. We'll introduce the high level steps of what the end-to-end ML lifecycle looks like and how different roles can collaborate to complete the ML project. Machine learning is a powerful tool to help solve different problems in your business. The article "Building your first machine learning model" gives you basic ideas of what it takes to build a machine learning model. In this article, we'll talk about what the end-to-end machine learning project lifecycle looks like in a real business.


Multi-Agent Cooperative Bidding Games for Multi-Objective Optimization in e-Commercial Sponsored Search

Guan, Ziyu, Wu, Hongchang, Cao, Qingyu, Liu, Hao, Zhao, Wei, Li, Sheng, Xu, Cai, Qiu, Guang, Xu, Jian, Zheng, Bo

arXiv.org Artificial Intelligence

Bid optimization for online advertising from single advertiser's perspective has been thoroughly investigated in both academic research and industrial practice. However, existing work typically assume competitors do not change their bids, i.e., the wining price is fixed, leading to poor performance of the derived solution. Although a few studies use multi-agent reinforcement learning to set up a cooperative game, they still suffer the following drawbacks: (1) They fail to avoid collusion solutions where all the advertisers involved in an auction collude to bid an extremely low price on purpose. (2) Previous works cannot well handle the underlying complex bidding environment, leading to poor model convergence. This problem could be amplified when handling multiple objectives of advertisers which are practical demands but not considered by previous work. In this paper, we propose a novel multi-objective cooperative bid optimization formulation called Multi-Agent Cooperative bidding Games (MACG). MACG sets up a carefully designed multi-objective optimization framework where different objectives of advertisers are incorporated. A global objective to maximize the overall profit of all advertisements is added in order to encourage better cooperation and also to protect self-bidding advertisers. To avoid collusion, we also introduce an extra platform revenue constraint. We analyze the optimal functional form of the bidding formula theoretically and design a policy network accordingly to generate auction-level bids. Then we design an efficient multi-agent evolutionary strategy for model optimization. Offline experiments and online A/B tests conducted on the Taobao platform indicate both single advertiser's objective and global profit have been significantly improved compared to state-of-art methods.


How YouTube is Recommending Your Next Video - KDnuggets

#artificialintelligence

In a recent paper [1] published by Google researchers and presented at RecSys 2019 (Copenhagen, Denmark) insight was provided in how their video platform Youtube recommends which videos to watch. In this blogpost I will try to summarise my findings after reading this paper. When users are watching videos on Youtube, a list of recommended videos are displayed which the user might like in a certain order. How to effectively and efficiently learn to reduce such biases is an open question. The described model in this paper focuses on the two main objectives.